14 research outputs found

    Interpretable Multi Labeled Bengali Toxic Comments Classification using Deep Learning

    Full text link
    This paper presents a deep learning-based pipeline for categorizing Bengali toxic comments, in which at first a binary classification model is used to determine whether a comment is toxic or not, and then a multi-label classifier is employed to determine which toxicity type the comment belongs to. For this purpose, we have prepared a manually labeled dataset consisting of 16,073 instances among which 8,488 are Toxic and any toxic comment may correspond to one or more of the six toxic categories - vulgar, hate, religious, threat, troll, and insult simultaneously. Long Short Term Memory (LSTM) with BERT Embedding achieved 89.42% accuracy for the binary classification task while as a multi-label classifier, a combination of Convolutional Neural Network and Bi-directional Long Short Term Memory (CNN-BiLSTM) with attention mechanism achieved 78.92% accuracy and 0.86 as weighted F1-score. To explain the predictions and interpret the word feature importance during classification by the proposed models, we utilized Local Interpretable Model-Agnostic Explanations (LIME) framework. We have made our dataset public and can be accessed at - https://github.com/deepu099cse/Multi-Labeled-Bengali-Toxic-Comments-Classificatio

    Huruf: An Application for Arabic Handwritten Character Recognition Using Deep Learning

    Full text link
    Handwriting Recognition has been a field of great interest in the Artificial Intelligence domain. Due to its broad use cases in real life, research has been conducted widely on it. Prominent work has been done in this field focusing mainly on Latin characters. However, the domain of Arabic handwritten character recognition is still relatively unexplored. The inherent cursive nature of the Arabic characters and variations in writing styles across individuals makes the task even more challenging. We identified some probable reasons behind this and proposed a lightweight Convolutional Neural Network-based architecture for recognizing Arabic characters and digits. The proposed pipeline consists of a total of 18 layers containing four layers each for convolution, pooling, batch normalization, dropout, and finally one Global average pooling and a Dense layer. Furthermore, we thoroughly investigated the different choices of hyperparameters such as the choice of the optimizer, kernel initializer, activation function, etc. Evaluating the proposed architecture on the publicly available 'Arabic Handwritten Character Dataset (AHCD)' and 'Modified Arabic handwritten digits Database (MadBase)' datasets, the proposed model respectively achieved an accuracy of 96.93% and 99.35% which is comparable to the state-of-the-art and makes it a suitable solution for real-life end-level applications.Comment: Accepted in 25th ICCIT (6 pages, 4 tables, 4 figures

    Effect of genotype on proximate composition and biological yield of maize (Zea mays L.)

    Get PDF
    An experiment was conducted to study the proximate composition of five released maize varieties (Zea mays L.) of Bangladesh Agricultural Research Institute (BARI), which was popularly growing in Bangladesh namely BHM-5, BHM-8, BHM-13, BHM-15, and Barnali. There was none a single variety performed best in all nutrient parameters. Among these maize varieties, the highest grain weight of 100 seeds, and yield was found in BHM-15 (32.84g and 12.6 ton/ha). In the case of proximate analysis, the highest protein, ash, and fat content was recorded from BHM-15 (13.11%, 2.33%, and 5.44%), the highest carbohydrate content was recorded from BHM-13 (82.40%), and the highest amount of fiber was recorded from BHM-5 (2.07%). On the other hand, the lowest amount of carbohydrate and protein was recorded from BHM-15 (77.67%) and BHM-8 (10.96%), respectively. BHM-13 contained the lowest amount of fiber (1.24%) and fat (4.27%). Barnali and BHM-15 showed better performance for most of the minerals. The findings concluded that the different genotypes of maize differ substantially in their chemical and mineral compositions

    Performance Analysis of YOLO-based Architectures for Vehicle Detection from Traffic Images in Bangladesh

    Full text link
    The task of locating and classifying different types of vehicles has become a vital element in numerous applications of automation and intelligent systems ranging from traffic surveillance to vehicle identification and many more. In recent times, Deep Learning models have been dominating the field of vehicle detection. Yet, Bangladeshi vehicle detection has remained a relatively unexplored area. One of the main goals of vehicle detection is its real-time application, where `You Only Look Once' (YOLO) models have proven to be the most effective architecture. In this work, intending to find the best-suited YOLO architecture for fast and accurate vehicle detection from traffic images in Bangladesh, we have conducted a performance analysis of different variants of the YOLO-based architectures such as YOLOV3, YOLOV5s, and YOLOV5x. The models were trained on a dataset containing 7390 images belonging to 21 types of vehicles comprising samples from the DhakaAI dataset, the Poribohon-BD dataset, and our self-collected images. After thorough quantitative and qualitative analysis, we found the YOLOV5x variant to be the best-suited model, performing better than YOLOv3 and YOLOv5s models respectively by 7 & 4 percent in mAP, and 12 & 8.5 percent in terms of Accuracy.Comment: Accepted in 25th ICCIT (6 pages, 5 figures, 1 table

    Facial Expression Recognition under Difficult Conditions: A Comprehensive Study on Edge Directional Texture Patterns

    No full text
    In recent years, research in automated facial expression recognition has attained significant attention for its potential applicability in human-computer interaction, surveillance systems, animation, and consumer electronics. However, recognition in uncontrolled environments under the presence of illumination and pose variations, low-resolution video, occlusion, and random noise is still a challenging research problem. In this paper, we investigate recognition of facial expression in difficult conditions by means of an effective facial feature descriptor, namely the directional ternary pattern (DTP). Given a face image, the DTP operator describes the facial feature by quantizing the eight-directional edge response values, capturing essential texture properties, such as presence of edges, corners, points, lines, etc. We also present an enhancement of the basic DTP encoding method, namely the compressed DTP (cDTP) that can describe the local texture more effectively with fewer features. The recognition performances of the proposed DTP and cDTP descriptors are evaluated using the Cohn-Kanade (CK) and the Japanese female facial expression (JAFFE) database. In our experiments, we simulate difficult conditions using original database images with lighting variations, low-resolution images obtained by down-sampling the original, and images corrupted with Gaussian noise. In all cases, the proposed method outperforms some of the well-known face feature descriptors

    GaitGCN++: Improving GCN-based gait recognition with part-wise attention and DropGraph

    No full text
    Gait recognition is becoming one of the promising methods for biometric authentication owing to its self-effacing nature. Contemporary approaches of joint position-based gait recognition generally model gait features using spatio-temporal graphs which are often prone to overfitting. To incorporate long-range relationships among joints, these methods utilize multi-scale operators. However, they fail to provide equal importance to all joint combinations resulting in an incomplete realization of long-range relationships between joints and important body parts. Furthermore, only considering joint coordinates may fail to capture discriminatory information provided by the bone structures and motion. In this work, a novel multi-scale graph convolution approach, namely ‘GaitGCN++’, is proposed, which utilizes joint and bone information from individual frames and joint-motion data from consecutive frames providing a comprehensive understanding of gait. An efficient hop-extraction technique is utilized to understand the relationship between closer and further joints while avoiding redundant dependencies. Additionally, traditional graph convolution is enhanced by leveraging the ‘DropGraph’ regularization technique to avoid overfitting and the ‘Part-wise Attention’ to identify the most important body parts over the gait sequence. On the benchmark gait recognition dataset CASIA-B and GREW, we outperform the state-of-the-art in diversified and challenging scenarios

    Less is More: Lighter and Faster Deep Neural Architecture for Tomato Leaf Disease Classification

    Full text link
    To ensure global food security and the overall profit of stakeholders, the importance of correctly detecting and classifying plant diseases is paramount. In this connection, the emergence of deep learning-based image classification has introduced a substantial number of solutions. However, the applicability of these solutions in low-end devices requires fast, accurate, and computationally inexpensive systems. This work proposes a lightweight transfer learning-based approach for detecting diseases from tomato leaves. It utilizes an effective preprocessing method to enhance the leaf images with illumination correction for improved classification. Our system extracts features using a combined model consisting of a pretrained MobileNetV2 architecture and a classifier network for effective prediction. Traditional augmentation approaches are replaced by runtime augmentation to avoid data leakage and address the class imbalance issue. Evaluation on tomato leaf images from the PlantVillage dataset shows that the proposed architecture achieves 99.30% accuracy with a model size of 9.60MB and 4.87M floating-point operations, making it a suitable choice for real-life applications in low-end devices. Our codes and models are available at https://github.com/redwankarimsony/project-tomato.Comment: 18 pages, 13 figures, 5 tables, Accepted in IEEE Acces
    corecore